Massively Parallel Polar Decomposition on Distributed-memory Systems
نویسندگان
چکیده
منابع مشابه
Parallel Mesh Partitioning on Distributed Memory Systems
We discuss the problem of deriving parallel mesh partitioning algorithms for mapping unstructured meshes to parallel computers. In itself this raises a paradox – we seek to find a high quality partition of the mesh, but to compute it in parallel we require a partition of the mesh. In fact, we overcome this difficulty by deriving an optimisation strategy which can find a high quality partition e...
متن کاملAutomatic Data and Computation Decomposition on Distributed Memory Parallel Computers1
On shared memory parallel computers (SMPCs) it is natural to focus on decomposing the computation (mainly by distributing the iterations of the nested Do-Loops). In contrast, on distributed memory parallel computers (DMPCs) the decomposition of computation and the distribution of data must both be handled—in order to balance the computation load and to minimize the migration of data. We propose...
متن کاملDesign Principle of Massively Parallel Distributed-Memory Multiprocessor Architecture
In this papery we discuss the design principles of massively parallel distributed-memory multiprocessor architecture and propose the Datarol-H architecture. We present the architecture of the massively parallel Datarol-H machine and show a Datarol-H processor design, including communication protocol and handling mechanisms of remote memory access and remote process/procedure invocation. Last, w...
متن کاملA Design Principle of Massively Parallel Distributed-Memory Multiprocessor Architecture
Distributed-memory multiprocessor architecture is essential in developing massively parallel machines. One of the most important design issues in such a distributed-memory multiprocessor architecture is a latency problem which is caused by remote procedure invocation and remote memory access. Remote memory access and remote procedure invocation occur so often in massively parallel execution, an...
متن کاملComparing and Improving Centralized and Distributed Techniques for Coordinating Massively Parallel Shared-Memory Systems
Two complementary approaches have been proposed to achieve high performance inter-process coordination on highly parallel shared-memory systems. Gottlieb et. al. introduced the technique of combining concurrent memory references, thereby reducing hot spot contention and enabling the “bottleneck-free” execution of algorithms referencing a small number of shared variables. MellorCrummey and Scott...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ACM Transactions on Parallel Computing
سال: 2019
ISSN: 2329-4949,2329-4957
DOI: 10.1145/3328723